- Home
- Search Results
- Page 1 of 1
Search for: All records
-
Total Resources3
- Resource Type
-
0001000002000000
- More
- Availability
-
30
- Author / Contributor
- Filter by Author / Creator
-
-
Sun, Runhan (3)
-
Dixon, Warren E. (2)
-
Bell, Zachary I. (1)
-
Bhasin, Shubhendu (1)
-
Chen, Hsi-Yuan (1)
-
Chowdhary, Girish (1)
-
Greene, Max L. (1)
-
Le, Duc M. (1)
-
Li, Xiaolin (1)
-
Lu, Fan (1)
-
Ma, Xiyao (1)
-
Patil, Omkar Sudhir (1)
-
Pu, George (1)
-
Zhou, Yanlin (1)
-
#Tyler Phillips, Kenneth E. (0)
-
#Willis, Ciara (0)
-
& Abreu-Ramos, E. D. (0)
-
& Abramson, C. I. (0)
-
& Abreu-Ramos, E. D. (0)
-
& Adams, S.G. (0)
-
- Filter by Editor
-
-
& Spizer, S. M. (0)
-
& . Spizer, S. (0)
-
& Ahn, J. (0)
-
& Bateiha, S. (0)
-
& Bosch, N. (0)
-
& Brennan K. (0)
-
& Brennan, K. (0)
-
& Chen, B. (0)
-
& Chen, Bodong (0)
-
& Drown, S. (0)
-
& Ferretti, F. (0)
-
& Higgins, A. (0)
-
& J. Peters (0)
-
& Kali, Y. (0)
-
& Ruiz-Arias, P.M. (0)
-
& S. Spitzer (0)
-
& Sahin. I. (0)
-
& Spitzer, S. (0)
-
& Spitzer, S.M. (0)
-
(submitted - in Review for IEEE ICASSP-2024) (0)
-
-
Have feedback or suggestions for a way to improve these results?
!
Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher.
Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?
Some links on this page may take you to non-federal websites. Their policies may differ from this site.
-
Sun, Runhan; Greene, Max L.; Le, Duc M.; Bell, Zachary I.; Chowdhary, Girish; Dixon, Warren E. (, IEEE Control Systems Letters)
-
Zhou, Yanlin; Lu, Fan; Pu, George; Ma, Xiyao; Sun, Runhan; Chen, Hsi-Yuan; Li, Xiaolin (, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS))We propose a deep reinforcement learning (DRL) methodology for the tracking, obstacle avoidance, and formation control of nonholonomic robots. By separating vision-based control into a perception module and a controller module, we can train a DRL agent without sophisticated physics or 3D modeling. In addition, the modular framework averts daunting retrains of an image-to-action end-to-end neural network, and provides flexibility in transferring the controller to different robots. First, we train a convolutional neural network (CNN) to accurately localize in an indoor setting with dynamic foreground/background. Then, we design a new DRL algorithm named Momentum Policy Gradient (MPG) for continuous control tasks and prove its convergence. We also show that MPG is robust at tracking varying leader movements and can naturally be extended to problems of formation control. Leveraging reward shaping, features such as collision and obstacle avoidance can be easily integrated into a DRL controller.more » « less
An official website of the United States government
